Frequently Asked Questions (FAQ)
How Can I Explain This to You? An Empirical Study of Deep Neural Network Explanation Methods
Explaining the inner workings of deep neural network models have received considerable attention in recent years. Researchers have attempted to provide human parseable explanations justifying why a model performed a specific classification. Although many of these toolkits are available for use, it is unclear which style of explanation is preferred by end-users, thereby demanding investigation. We performed a cross-analysis Amazon Mechanical Turk study comparing the popular state-of-the-art explanation methods to empirically determine which are better in explaining model decisions. The participants were asked to compare explanation methods across applications spanning image, text, audio, and sensory domains. Among the surveyed methods, explanation-by-example was preferred in all domains except text sentiment classification, where LIME's method of annotating input text was preferred. We highlight qualitative aspects of employing the studied explainability methods and conclude with implications for researchers and engineers that seek to incorporate explanations into user-facing deployments.
MAFA: A multi-agent framework for annotation
Hegazy, Mahmood, Rodrigues, Aaron, Naeem, Azzam
Modern consumer banking applications require accurate and efficient retrieval of information in response to user queries. Mapping user utterances to the most relevant Frequently Asked Questions (FAQs) is a crucial component of these systems. Traditional approaches often rely on a single model or technique, which may not capture the nuances of diverse user inquiries. In this paper, we introduce a multi-agent framework for FAQ annotation that combines multiple specialized agents with different approaches and a judge agent that reranks candidates to produce optimal results. Our agents utilize a structured reasoning approach inspired by Attentive Reasoning Queries (ARQs), which guides them through systematic reasoning steps using targeted, task-specific JSON queries. Our framework features a few-shot example strategy, where each agent receives different few-shots, enhancing ensemble diversity and coverage of the query space. We evaluate our framework on a real-world major bank dataset as well as public benchmark datasets (LCQMC and FiQA), demonstrating significant improvements over single-agent approaches across multiple metrics, including a 14% increase in Top-1 accuracy, an 18% increase in Top-5 accuracy, and a 12% improvement in Mean Reciprocal Rank on our dataset, and similar gains on public benchmarks when compared with traditional and single-agent annotation techniques. Our framework is particularly effective at handling ambiguous queries, making it well-suited for deployment in production banking applications while showing strong generalization capabilities across different domains and languages.
- North America > United States > New Mexico > Santa Fe County > Santa Fe (0.04)
- North America > Mexico > Mexico City > Mexico City (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
- (2 more...)
- Frequently Asked Questions (FAQ) (1.00)
- Research Report (0.64)
- Banking & Finance (1.00)
- Information Technology > Security & Privacy (0.46)
Japan group to launch AI service for saury size predictions
Saury catches from August to the end of September this year totaled about 28,500 tons -- a 2.4-fold increase from the same period last year. The Japan Fisheries Information Service Center will start a service next fishing season that shows expected fishing grounds for saury by size class based on analysis using artificial intelligence technology. The Tokyo-based group of fisheries organizations provides information on fishing and ocean conditions. Since 2020, the group provides its predictions of likely saury fishing spots using AI, based on seawater temperature changes and past fishing records. The accuracy of the predictions has improved year after year.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.29)
- Asia > Japan > Honshū > Kansai > Osaka Prefecture > Osaka (0.07)
- North America > United States (0.05)
- (4 more...)
- Media > News (0.30)
- Leisure & Entertainment (0.30)
- Information Technology > Communications > Social Media (0.75)
- Information Technology > Artificial Intelligence > Applied AI (0.55)
Review for NeurIPS paper: How Can I Explain This to You? An Empirical Study of Deep Neural Network Explanation Methods
Weaknesses: The term'unified' should be revised as the paper addresses a partial unification. For instance, the unified framework does not take into account a closed loop between the DNN and the explanation method (the explanation method can be itself another DNN interacting in a double sense with the prediction DNN) or other two-stage adaptive networks [1], [2]. In addition, an alternative to example based explanation is'opening the black box' in terms of intra-layer and inter-layer statistical properties of a DNN [3]: these may be enough to explain lack of generality (and thus absence of recommendation) of a given network depending on the input available data and the classification paradigm considered. Thus, a positioning must be provided with respect to the above issues in order to make the paper more informative with respect to the literature. The weak spots of the analysis are twofold.
Review for NeurIPS paper: How Can I Explain This to You? An Empirical Study of Deep Neural Network Explanation Methods
The paper is an empirical study on the types of explanations preferred by users using AMT. All reviewers found that problem was important and found the study interesting. However, one reviewer argued that while this was a good first step, it does not address the fact evaluating explanations is an ill-posed problem. Three reviewers found that this study is interesting enough to the NeurISP community even as a first step.
Do Voters Get the Information They Want? Understanding Authentic Voter FAQs in the US and How to Improve for Informed Electoral Participation
Rawte, Vipula, Scott, Deja N, Kumar, Gaurav, Juneja, Aishneet, Yaddanapalli, Bharat Sowrya, Srivastava, Biplav
Accurate information is crucial for democracy as it empowers voters to make informed decisions about their representatives and keeping them accountable. In the US, state election commissions (SECs), often required by law, are the primary providers of Frequently Asked Questions (FAQs) to voters, and secondary sources like non-profits such as League of Women Voters (LWV) try to complement their information shortfall. However, surprisingly, to the best of our knowledge, there is neither a single source with comprehensive FAQs nor a study analyzing the data at national level to identify current practices and ways to improve the status quo. This paper addresses it by providing the {\bf first dataset on Voter FAQs covering all the US states}. Second, we introduce metrics for FAQ information quality (FIQ) with respect to questions, answers, and answers to corresponding questions. Third, we use FIQs to analyze US FAQs to identify leading, mainstream and lagging content practices and corresponding states. Finally, we identify what states across the spectrum can do to improve FAQ quality and thus, the overall information ecosystem. Across all 50 U.S. states, 12% were identified as leaders and 8% as laggards for FIQS\textsubscript{voter}, while 14% were leaders and 12% laggards for FIQS\textsubscript{developer}.
- North America > United States > South Carolina (0.04)
- North America > United States > Michigan (0.04)
- North America > United States > Wisconsin (0.04)
- (47 more...)
Agentic AI-Driven Technical Troubleshooting for Enterprise Systems: A Novel Weighted Retrieval-Augmented Generation Paradigm
Technical troubleshooting in enterprise environments often involves navigating diverse, heterogeneous data sources to resolve complex issues effectively. This paper presents a novel agentic AI solution built on a Weighted Retrieval-Augmented Generation (RAG) Framework tailored for enterprise technical troubleshooting. By dynamically weighting retrieval sources such as product manuals, internal knowledge bases, FAQs, and troubleshooting guides based on query context, the framework prioritizes the most relevant data. For instance, it gives precedence to product manuals for SKU-specific queries while incorporating general FAQs for broader issues. The system employs FAISS for efficient dense vector search, coupled with a dynamic aggregation mechanism to seamlessly integrate results from multiple sources. A Llama-based self-evaluator ensures the contextual accuracy and confidence of the generated responses before delivering them. This iterative cycle of retrieval and validation enhances precision, diversity, and reliability in response generation. Preliminary evaluations on large enterprise datasets demonstrate the framework's efficacy in improving troubleshooting accuracy, reducing resolution times, and adapting to varied technical challenges. Future research aims to enhance the framework by integrating advanced conversational AI capabilities, enabling more interactive and intuitive troubleshooting experiences. Efforts will also focus on refining the dynamic weighting mechanism through reinforcement learning to further optimize the relevance and precision of retrieved information. By incorporating these advancements, the proposed framework is poised to evolve into a comprehensive, autonomous AI solution, redefining technical service workflows across enterprise settings.
- Research Report (0.64)
- Frequently Asked Questions (FAQ) (0.56)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Information Retrieval (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Beyond-RAG: Question Identification and Answer Generation in Real-Time Conversations
Agrawal, Garima, Gummuluri, Sashank, Spera, Cosimo
In customer contact centers, human agents often struggle with long average handling times (AHT) due to the need to manually interpret queries and retrieve relevant knowledge base (KB) articles. While retrieval augmented generation (RAG) systems using large language models (LLMs) have been widely adopted in industry to assist with such tasks, RAG faces challenges in real-time conversations, such as inaccurate query formulation and redundant retrieval of frequently asked questions (FAQs). To address these limitations, we propose a decision support system that can look beyond RAG by first identifying customer questions in real time. If the query matches an FAQ, the system retrieves the answer directly from the FAQ database; otherwise, it generates answers via RAG. Our approach reduces reliance on manual queries, providing responses to agents within 2 seconds. Deployed in AI-powered human-agent assist solution at Minerva CQ, this system improves efficiency, reduces AHT, and lowers operational costs. We also introduce an automated LLM-agentic workflow to identify FAQs from historical transcripts when no predefined FAQs exist.
- North America > United States > California > Santa Clara County > Sunnyvale (0.04)
- North America > United States > Arizona > Maricopa County > Tempe (0.04)
How Can I Explain This to You? An Empirical Study of Deep Neural Network Explanation Methods
Explaining the inner workings of deep neural network models have received considerable attention in recent years. Researchers have attempted to provide human parseable explanations justifying why a model performed a specific classification. Although many of these toolkits are available for use, it is unclear which style of explanation is preferred by end-users, thereby demanding investigation. We performed a cross-analysis Amazon Mechanical Turk study comparing the popular state-of-the-art explanation methods to empirically determine which are better in explaining model decisions. The participants were asked to compare explanation methods across applications spanning image, text, audio, and sensory domains.
Five questions and answers about artificial intelligence
Prieto, Alberto, Prieto, Beatriz
Rapid advances in Artificial Intelligence (AI) are generating much controversy in society, often without scientific basis. As occurred the development of other emerging technologies, such as the introduction of electricity in the early 20th century, AI causes both fascination and fear. Following the advice of the philosopher R.W. Emerson's advice'the knowledge is the antidote to fear', this paper seeks to contribute to the dissemination of knowledge about AI. To this end, it reflects on the following questions: the origins of AI, its possible future evolution, its ability to show feelings, the associated threats and dangers, and the concept of AI singularity Keywords: Artificial Intelligence (AI), Fourth Industrial Revolution, Beginnings of AI, Development of AI, Automatic learning, Machine learning, Feelings in AI, Dangers of AI, Advantages of AI, Singularity of AI, Superintelligence, Frictionless Reproducibility (FR), Large Language Models, General AI (GAI), Intelligence, GPT Chat.
- North America > United States > Pennsylvania (0.04)
- North America > United States > New York (0.04)
- Europe > Spain > Catalonia (0.04)
- Europe > Spain > Andalusia > Granada Province > Granada (0.04)
- Frequently Asked Questions (FAQ) (0.41)
- Research Report (0.40)
- Leisure & Entertainment > Games > Chess (1.00)
- Law (0.93)
- Government (0.93)
- (2 more...)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Issues > Social & Ethical Issues (1.00)
- Information Technology > Artificial Intelligence > Cognitive Science (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)